On Bochner's and Polya's Characterizations of Positive-Definite Kernels and the Respective Random Feature Maps
نویسندگان
چکیده
Positive-definite kernel functions are fundamental elements of kernel methods and Gaussian processes. A well-known construction of such functions comes from Bochner’s characterization, which connects a positive-definite function with a probability distribution. Another construction, which appears to have attracted less attention, is Polya’s criterion that characterizes a subset of these functions. In this paper, we study the latter characterization and derive a number of novel kernels little known previously. In the context of large-scale kernel machines, Rahimi and Recht (2007) proposed a random feature map (random Fourier) that approximates a kernel function, through independent sampling of the probability distribution in Bochner’s characterization. The authors also suggested another feature map (random binning), which, although not explicitly stated, comes from Polya’s characterization. We show that with the same number of random samples, the random binning map results in an Euclidean inner product closer to the kernel than does the random Fourier map. The superiority of the random binning map is confirmed empirically through regressions and classifications in the reproducing kernel Hilbert space.
منابع مشابه
Operator-Valued Bochner Theorem, Fourier Feature Maps for Operator-Valued Kernels, and Vector-Valued Learning
This paper presents a framework for computing random operator-valued feature maps for operator-valued positive definite kernels. This is a generalization of the random Fourier features for scalar-valued kernels to the operator-valued case. Our general setting is that of operator-valued kernels corresponding to RKHS of functions with values in a Hilbert space. We show that in general, for a give...
متن کاملRandom Feature Maps for Dot Product Kernels Supplementary Material
This document contains detailed proofs of theorems stated in the main article entitled Random Feature Maps for Dot Product Kernels. 1 Proof of Theorem 1 We first recollect Schoenberg’s result in its original form Theorem 1 (Schoenberg (1942), Theorem 2). A function f : [−1, 1]→ R constitutes a positive definite kernel K : S∞ × S∞ → R, K : (x,y) 7→ f(〈x,y〉) iff f is an analytic function admittin...
متن کاملMercer's Theorem, Feature Maps, and Smoothing
We study Mercer’s theorem and feature maps for several positive definite kernels that are widely used in practice. The smoothing properties of these kernels will also be explored.
متن کاملAn efficient method for cloud detection based on the feature-level fusion of Landsat-8 OLI spectral bands in deep convolutional neural network
Cloud segmentation is a critical pre-processing step for any multi-spectral satellite image application. In particular, disaster-related applications e.g., flood monitoring or rapid damage mapping, which are highly time and data-critical, require methods that produce accurate cloud masks in a short time while being able to adapt to large variations in the target domain (induced by atmospheric c...
متن کاملA Geometry Preserving Kernel over Riemannian Manifolds
Abstract- Kernel trick and projection to tangent spaces are two choices for linearizing the data points lying on Riemannian manifolds. These approaches are used to provide the prerequisites for applying standard machine learning methods on Riemannian manifolds. Classical kernels implicitly project data to high dimensional feature space without considering the intrinsic geometry of data points. ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
- CoRR
دوره abs/1610.08861 شماره
صفحات -
تاریخ انتشار 2016